Kernels on graphs have had limited options for node-level problems. To address this, we present a novel, generalized kernel for graphs with node feature data for semi-supervised learning. The kernel is derived from a regularization framework by treating the graph and feature data as two Hilbert spaces. We also show how numerous kernel-based models on graphs are instances of our design. A kernel defined this way has transductive properties, and this leads to improved ability to learn on fewer training points, as well as better handling of highly non-Euclidean data. We demonstrate these advantages using synthetic data where the distribution of the whole graph can inform the pattern of the labels. Finally, by utilizing a flexible polynomial of the graph Laplacian within the kernel, the model also performed effectively in semi-supervised classification on graphs of various levels of homophily.
translated by 谷歌翻译
我们提出了一种雷达惯性内径测量的方法,其使用连续时间框架来熔断来自多个汽车雷达的熔丝测量和惯性测量单元(IMU)。不利的天气条件对雷达传感器的操作性能不同,与相机和激光器传感器不同,对雷达传感器的操作性能没有显着影响。雷达在这种情况下的鲁棒性和乘客车辆雷达的普遍普遍激励我们来看看雷达用于自我运动估计。连续时间轨迹表示不仅应用于实现异构和异步多传感器融合的框架,还应用于通过能够计算封闭形式的姿势及其衍生物来实现高效优化,并且在任何特定时间沿着弹道。我们将我们的连续时间估计与来自离散时间雷达 - 惯性内径型方法的方法进行比较,并表明我们的连续时间方法优于离散时间方法。据我们所知,这是第一次将连续时间框架应用于雷达惯性内径术。
translated by 谷歌翻译
图形神经网络(GNN)在各种图挖掘任务中取得了巨大的成功。但是,当GNN堆叠着许多层时,总是观察到急剧性能降解。结果,大多数GNN仅具有浅层建筑,这限制了它们的表现力和对深社区的开发。最近的研究将深度GNN的性能降低归因于\ textit {过度平滑}的问题。在本文中,我们将传统的图形卷积操作分为两个独立操作:\ textit {passagation}(\ textbf {p})和\ textit {transformation}(\ textbf {t})。可以分为传播深度($ d_p $)和转换深度($ d_t $)。通过广泛的实验,我们发现深度GNNS性能下降的主要原因是\ textit {model dygradation}问题是由大$ d_t $而不是\ textit {过度平滑}问题引起的,主要是由大$ d_p $引起的。 。此外,我们提出\ textIt {自适应初始残留}(air),一个与各种GNN架构兼容的插件模块,以减轻\ textit {model {model dradation degradation}问题和\ textit {textit {过度敏感}问题同时。六个现实世界数据集的实验结果表明,配备空气的GNN胜过大多数具有浅层建筑的GNN,这是由于大型$ d_p $和$ d_t $的好处,而与空气相关的时间成本则可以忽略。
translated by 谷歌翻译
图形神经网络(GNN)在许多基于图的应用程序中取得了巨大成功。但是,巨大的尺寸和高稀疏度的图表阻碍了其在工业场景下的应用。尽管为大规模图提出了一些可扩展的GNN,但它们为每个节点采用固定的$ k $ hop邻域,因此在稀疏区域内采用大型繁殖深度时面临过度光滑的问题。为了解决上述问题,我们提出了一种新的GNN体系结构 - 图形注意多层感知器(GAMLP),该架构可以捕获不同图形知识范围之间的基本相关性。我们已经与天使平台部署了GAMLP,并进一步评估了现实世界数据集和大规模工业数据集的GAMLP。这14个图数据集的广泛实验表明,GAMLP在享有高可扩展性和效率的同时,达到了最先进的性能。具体来说,在我们的大规模腾讯视频数据集上的预测准确性方面,它的表现优于1.3 \%,同时达到了高达$ 50 \ times $ triending的速度。此外,它在开放图基准的最大同质和异质图(即OGBN-PAPERS100M和OGBN-MAG)的排行榜上排名第一。
translated by 谷歌翻译
近年来,神经网络在各个领域中表现出强大的力量,它也带来了越来越多的安全威胁。基于神经网络模型的STEGOMALWARE是代表性的。以前的研究初步证明通过突出神经网络模型中的恶意软件来启动恶意攻击的可行性。然而,现有的作品没有表明,由于恶意软件嵌入率低,模型性能降低以及额外的努力,这种新兴威胁在现实世界攻击中是实际的攻击。因此,我们预测一个称为evilmodel的改进的斯佩塔科。在分析神经网络模型的结构的基础上,我们将二进制形成恶意软件作为其参数嵌入神经网络模型,并提出了三种新的恶意软件嵌入技术,即MSB保留,快速替换和半替换。通过结婚19个恶意软件样本和10个流行的神经网络模型,我们构建了550个恶意软件嵌入式模型,并在想象中数据集中分析了这些模型的性能。实验结果表明,半取代几乎完美地表现出,恶意软件嵌入率为48.52%,没有模型性能下降或额外的努力。考虑到一系列因素,我们提出了一种定量算法来评估不同的嵌入方法。评估结果表明,邪恶的模型与经典的斯托图尼特有多高。此外,我们开展案例研究,以触发真实世界的情景中的邪恶模型。要深入了解所提出的恶意软件嵌入技术,我们还研究了神经网络结构,层和参数大小对恶意软件嵌入容量和嵌入式模型精度的影响。我们还提供了一些可能的对策来捍卫邪恶的模型。我们希望这项工作能够全面了解这种新的AI动力威胁,并建议提前辩护。
translated by 谷歌翻译
Graph神经网络(GNN)最近在许多基于图的应用程序中都实现了最先进的性能。尽管具有很高的表现力,但他们通常需要在多个培训时期进行昂贵的递归邻里扩展,并面临可伸缩性问题。此外,它们中的大多数是不灵活的,因为它们仅限于固定跳跃社区,并且对不同节点的实际接受场需求不敏感。我们通过引入可扩展且灵活的图表多层感知器(GAMLP)来规避这些限制。随着非线性转化和特征传播的分离,GAMLP通过以预先计算的方式执行传播程序来显着提高可伸缩性和效率。有了三个原则的接受场注意力,GAMLP中的每个节点都具有灵活性和适应性,以利用接收场的不同尺寸的传播特征。我们对三个大型开放图基准(例如OGBN-PAPERS100M,OGBN产品和OGBN-MAG)进行了广泛的评估,这表明GAMLP不仅可以实现前面的性能,而且还提供了较高的可扩展性和效率。
translated by 谷歌翻译
命令和控制(C&C)在攻击中很重要。它将命令从攻击者传输到受损的主机中的恶意软件。目前,一些攻击者在C&C任务中使用在线社交网络(OSN)。 OSN的C&C中有两个主要问题。首先,恶意软件找到攻击者的过程是可逆的。如果防御者分析了恶意软件样本,则在发布命令之前将暴露攻击者。其次,以普通或加密形式的命令被OSN视为异常内容,这会引起异常并触发攻击者的限制。防御者暴露后可以限制攻击者。在这项工作中,我们建议在OSN上使用AI驱动的C&C DEEPC2来解决这些问题。对于可逆的硬编码,恶意软件使用神经网络模型找到了攻击者。攻击者的头像被转换为​​一批特征向量,并且防御者无法使用模型和特征向量提前恢复头像。为了求解OSN上的异常内容,哈希碰撞和文本数据扩展用于将命令嵌入正常内容中。 Twitter上的实验表明,可以有效地生成命令包裹的推文。恶意软件可以在OSN上秘密地找到攻击者。安全分析表明,很难提前恢复攻击者的标识符。
translated by 谷歌翻译
We propose a distributionally robust return-risk model for Markov decision processes (MDPs) under risk and reward ambiguity. The proposed model optimizes the weighted average of mean and percentile performances, and it covers the distributionally robust MDPs and the distributionally robust chance-constrained MDPs (both under reward ambiguity) as special cases. By considering that the unknown reward distribution lies in a Wasserstein ambiguity set, we derive the tractable reformulation for our model. In particular, we show that that the return-risk model can also account for risk from uncertain transition kernel when one only seeks deterministic policies, and that a distributionally robust MDP under the percentile criterion can be reformulated as its nominal counterpart at an adjusted risk level. A scalable first-order algorithm is designed to solve large-scale problems, and we demonstrate the advantages of our proposed model and algorithm through numerical experiments.
translated by 谷歌翻译
We introduce a new tool for stochastic convex optimization (SCO): a Reweighted Stochastic Query (ReSQue) estimator for the gradient of a function convolved with a (Gaussian) probability density. Combining ReSQue with recent advances in ball oracle acceleration [CJJJLST20, ACJJS21], we develop algorithms achieving state-of-the-art complexities for SCO in parallel and private settings. For a SCO objective constrained to the unit ball in $\mathbb{R}^d$, we obtain the following results (up to polylogarithmic factors). We give a parallel algorithm obtaining optimization error $\epsilon_{\text{opt}}$ with $d^{1/3}\epsilon_{\text{opt}}^{-2/3}$ gradient oracle query depth and $d^{1/3}\epsilon_{\text{opt}}^{-2/3} + \epsilon_{\text{opt}}^{-2}$ gradient queries in total, assuming access to a bounded-variance stochastic gradient estimator. For $\epsilon_{\text{opt}} \in [d^{-1}, d^{-1/4}]$, our algorithm matches the state-of-the-art oracle depth of [BJLLS19] while maintaining the optimal total work of stochastic gradient descent. We give an $(\epsilon_{\text{dp}}, \delta)$-differentially private algorithm which, given $n$ samples of Lipschitz loss functions, obtains near-optimal optimization error and makes $\min(n, n^2\epsilon_{\text{dp}}^2 d^{-1}) + \min(n^{4/3}\epsilon_{\text{dp}}^{1/3}, (nd)^{2/3}\epsilon_{\text{dp}}^{-1})$ queries to the gradients of these functions. In the regime $d \le n \epsilon_{\text{dp}}^{2}$, where privacy comes at no cost in terms of the optimal loss up to constants, our algorithm uses $n + (nd)^{2/3}\epsilon_{\text{dp}}^{-1}$ queries and improves recent advancements of [KLL21, AFKT21]. In the moderately low-dimensional setting $d \le \sqrt n \epsilon_{\text{dp}}^{3/2}$, our query complexity is near-linear.
translated by 谷歌翻译
New architecture GPUs like A100 are now equipped with multi-instance GPU (MIG) technology, which allows the GPU to be partitioned into multiple small, isolated instances. This technology provides more flexibility for users to support both deep learning training and inference workloads, but efficiently utilizing it can still be challenging. The vision of this paper is to provide a more comprehensive and practical benchmark study for MIG in order to eliminate the need for tedious manual benchmarking and tuning efforts. To achieve this vision, the paper presents MIGPerf, an open-source tool that streamlines the benchmark study for MIG. Using MIGPerf, the authors conduct a series of experiments, including deep learning training and inference characterization on MIG, GPU sharing characterization, and framework compatibility with MIG. The results of these experiments provide new insights and guidance for users to effectively employ MIG, and lay the foundation for further research on the orchestration of hybrid training and inference workloads on MIGs. The code and results are released on https://github.com/MLSysOps/MIGProfiler. This work is still in progress and more results will be published soon.
translated by 谷歌翻译